Goto

Collaborating Authors

 unstructured demonstration


Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets

Neural Information Processing Systems

Imitation learning has traditionally been applied to learn a single task from demonstrations thereof. The requirement of structured and isolated demonstrations limits the scalability of imitation learning approaches as they are difficult to apply to real-world scenarios, where robots have to be able to execute a multitude of tasks. In this paper, we propose a multi-modal imitation learning framework that is able to segment and imitate skills from unlabelled and unstructured demonstrations by learning skill segmentation and imitation learning jointly. The extensive simulation results indicate that our method can efficiently separate the demonstrations into individual skills and learn to imitate them using a single multi-modal policy.


Reviews: Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets

Neural Information Processing Systems

The paper describes a new learning model able to discover'intentions' from expert policies by using an imitation learning framework. The idea is mainly based on the GAIL model which aims at learning by imitation a policy using a GAN approach. The main difference in the article is that the learned policy is, in fact, a mixture of sub-policies, each sub-policy aiming at automatically matching a particular intention in the expert behavior. The GAIL algorithm is thus derived with this mixture, resulting in an effective learning technique. Another approach is also proposed where the intention will be captured through a latent vector by derivating the InfoGAN algorithm for this particular case.


Multi-Modal Imitation Learning from Unstructured Demonstrations using Generative Adversarial Nets

Hausman, Karol, Chebotar, Yevgen, Schaal, Stefan, Sukhatme, Gaurav, Lim, Joseph J.

Neural Information Processing Systems

Imitation learning has traditionally been applied to learn a single task from demonstrations thereof. The requirement of structured and isolated demonstrations limits the scalability of imitation learning approaches as they are difficult to apply to real-world scenarios, where robots have to be able to execute a multitude of tasks. In this paper, we propose a multi-modal imitation learning framework that is able to segment and imitate skills from unlabelled and unstructured demonstrations by learning skill segmentation and imitation learning jointly. The extensive simulation results indicate that our method can efficiently separate the demonstrations into individual skills and learn to imitate them using a single multi-modal policy. Papers published at the Neural Information Processing Systems Conference.


An Integrated System for Learning Multi-Step Robotic Tasks from Unstructured Demonstrations

Niekum, Scott (University of Massachusetts Amherst)

AAAI Conferences

We present an integrated system for segmenting demonstrations, recognizing repeated skills, and generalizing multi-step tasks from unstructured demonstrations.  This method combines recent work in Bayesian nonparametric statistics and learning from demonstration with perception using an RGB-D camera to generalize a multi-step task on the PR2 mobile manipulator.  We demonstrate the potential of our framework to learn a large library of skills over time and discuss how it might be improved with additional integration of components such as active learning, interactive feedback from humans, and more advanced perception. 


Complex Task Learning from Unstructured Demonstrations

Niekum, Scott (University of Massachusetts Amherst)

AAAI Conferences

Much work in learning from demonstration has focused on learning simple tasks from structured demonstrations that have a well-defined beginning and end. As we attempt to scale robot learning to increasingly complex tasks, it becomes intractable to learn task policies monolithically. Furthermore, it is desirable to be able to learn from natural, unstructured demonstrations, which are unsegmented, possibly incomplete, and may come from different tasks. We propose a three-part approach to designing a natural, scalable system that allows a robot to learn tasks of increasing complexity by automatically building and refining a library of skills over time. First, we describe a Bayesian nonparametric model that can segment unstructured demonstrations into appropriate numbers of component skills and recognize repeated skills across demonstrations and tasks. These skills can then be parameterized and generalized to new situations. Second, we propose to create a system that allows the user to provide unstructured corrections and feedback to the robot, without requiring any knowledge of the robot's underlying representation of the task or its component skills. Third, we propose to infer the user's intentions for each segmented skill and autonomously improve these skills using reinforcement learning. This approach will be applied to learn and generalize complex, multi-step tasks that are beyond the reach of current LfD methods, using the PR2 mobile manipulator as a testing platform.